232 research outputs found

    An Approximately Optimal Algorithm for Scheduling Phasor Data Transmissions in Smart Grid Networks

    Full text link
    In this paper, we devise a scheduling algorithm for ordering transmission of synchrophasor data from the substation to the control center in as short a time frame as possible, within the realtime hierarchical communications infrastructure in the electric grid. The problem is cast in the framework of the classic job scheduling with precedence constraints. The optimization setup comprises the number of phasor measurement units (PMUs) to be installed on the grid, a weight associated with each PMU, processing time at the control center for the PMUs, and precedence constraints between the PMUs. The solution to the PMU placement problem yields the optimum number of PMUs to be installed on the grid, while the processing times are picked uniformly at random from a predefined set. The weight associated with each PMU and the precedence constraints are both assumed known. The scheduling problem is provably NP-hard, so we resort to approximation algorithms which provide solutions that are suboptimal yet possessing polynomial time complexity. A lower bound on the optimal schedule is derived using branch and bound techniques, and its performance evaluated using standard IEEE test bus systems. The scheduling policy is power grid-centric, since it takes into account the electrical properties of the network under consideration.Comment: 8 pages, published in IEEE Transactions on Smart Grid, October 201

    Signal Reconstruction via H-infinity Sampled-Data Control Theory: Beyond the Shannon Paradigm

    Get PDF
    This paper presents a new method for signal reconstruction by leveraging sampled-data control theory. We formulate the signal reconstruction problem in terms of an analog performance optimization problem using a stable discrete-time filter. The proposed H-infinity performance criterion naturally takes intersample behavior into account, reflecting the energy distributions of the signal. We present methods for computing optimal solutions which are guaranteed to be stable and causal. Detailed comparisons to alternative methods are provided. We discuss some applications in sound and image reconstruction

    Online Algorithms for Dynamic Matching Markets in Power Distribution Systems

    Full text link
    This paper proposes online algorithms for dynamic matching markets in power distribution systems, which at any real-time operation instance decides about matching -- or delaying the supply of -- flexible loads with available renewable generation with the objective of maximizing the social welfare of the exchange in the system. More specifically, two online matching algorithms are proposed for the following generation-load scenarios: (i) when the mean of renewable generation is greater than the mean of the flexible load, and (ii) when the condition (i) is reversed. With the intuition that the performance of such algorithms degrades with increasing randomness of the supply and demand, two properties are proposed for assessing the performance of the algorithms. First property is convergence to optimality (CO) as the underlying randomness of renewable generation and customer loads goes to zero. The second property is deviation from optimality, is measured as a function of the standard deviation of the underlying randomness of renewable generation and customer loads. The algorithm proposed for the first scenario is shown to satisfy CO and a deviation from optimal that varies linearly with the variation in the standard deviation. But the same algorithm is shown to not satisfy CO for the second scenario. We then show that the algorithm proposed for the second scenario satisfies CO and a deviation from optimal that varies linearly with the variation in standard deviation plus an offset

    Decentralized control and periodic feedback

    Get PDF
    Cataloged from PDF version of article.The decentralized stabilization problem for linear, discretetime, periodically timevarying plants using periodic controllers is considered. The main tool used isl the technique of Uning a periodic system to a timeinvariant one via extensions of the input and output spaces. It is shown that a periodically time-varying system of fundamental period N can be stabilized by a decentralized periodic controller if and only if: 1) the system is stabilizable and detectable, and 2) the N-lifting of each complementary subsystem of identieally zero inpnt-ontput map is free of unstable input-output decoupling zeros. In the special case of N = 1, this yields and clarifies all the mr exisling results on decentralized stabilization of time-invariant plants by periodically time varying controllers

    Online Learning Robust Control of Nonlinear Dynamical Systems

    Full text link
    In this work we address the problem of the online robust control of nonlinear dynamical systems perturbed by disturbance. We study the problem of attenuation of the total cost over a duration TT in response to the disturbances. We consider the setting where the cost function (at a particular time) is a general continuous function and adversarial, the disturbance is adversarial and bounded at any point of time. Our goal is to design a controller that can learn and adapt to achieve a certain level of attenuation. We analyse two cases (i) when the system is known and (ii) when the system is unknown. We measure the performance of the controller by the deviation of the controller's cost for a sequence of cost functions with respect to an attenuation γ\gamma, RtpR^p_t. We propose an online controller and present guarantees for the metric RtpR^p_t when the maximum possible attenuation is given by γ\overline{\gamma}, which is a system constant. We show that when the controller has preview of the cost functions and the disturbances for a short duration of time and the system is known RTp(γ)=O(1)R^p_T(\gamma) = O(1) when γγc\gamma \geq \gamma_c, where γc=O(γ)\gamma_c = \mathcal{O}(\overline{\gamma}). We then show that when the system is unknown the proposed controller with a preview of the cost functions and the disturbances for a short horizon achieves RTp(γ)=O(N)+O(1)+O((TN)g(N))R^p_T(\gamma) = \mathcal{O}(N) + \mathcal{O}(1) + \mathcal{O}((T-N)g(N)), when γγc\gamma \geq \gamma_c, where g(N)g(N) is the accuracy of a given nonlinear estimator and NN is the duration of the initial estimation period. We also characterize the lower bound on the required prediction horizon for these guarantees to hold in terms of the system constants

    Online Learning for Incentive-Based Demand Response

    Full text link
    In this paper, we consider the problem of learning online to manage Demand Response (DR) resources. A typical DR mechanism requires the DR manager to assign a baseline to the participating consumer, where the baseline is an estimate of the counterfactual consumption of the consumer had it not been called to provide the DR service. A challenge in estimating baseline is the incentive the consumer has to inflate the baseline estimate. We consider the problem of learning online to estimate the baseline and to optimize the operating costs over a period of time under such incentives. We propose an online learning scheme that employs least-squares for estimation with a perturbation to the reward price (for the DR services or load curtailment) that is designed to balance the exploration and exploitation trade-off that arises with online learning. We show that, our proposed scheme is able to achieve a very low regret of O((logT)2)\mathcal{O}\left((\log{T})^2\right) with respect to the optimal operating cost over TT days of the DR program with full knowledge of the baseline, and is individually rational for the consumers to participate. Our scheme is significantly better than the averaging type approach, which only fetches O(T1/3)\mathcal{O}(T^{1/3}) regret

    Meta-Learning Guarantees for Online Receding Horizon Learning Control

    Full text link
    In this paper we provide provable regret guarantees for an online meta-learning receding horizon control algorithm in an iterative control setting. We consider the setting where, in each iteration the system to be controlled is a linear deterministic system that is different and unknown, the cost for the controller in an iteration is a general additive cost function and there are affine control input constraints. By analysing conditions under which sub-linear regret is achievable, we prove that the online receding horizon controller achieves a regret for the controller cost and constraint violation that are O~(T3/4)\tilde{O}(T^{3/4}) with respect to the best policy that satisfies the control input control constraints, when the preview of the cost functions is limited to an interval and the interval size is doubled from one to the next. We then show that the average of the regret for the controller cost and constraint violation with respect to the same policy vary as O~((1+1/N)T3/4)\tilde{O}((1+1/\sqrt{N})T^{3/4}) with the number of iterations NN, under the same setting.Comment: arXiv admin note: substantial text overlap with arXiv:2008.13265, arXiv:2010.0726
    corecore